Adam in Private: Secure and Fast Training of Deep Neural Networks with Adaptive Moment Estimation
نویسندگان
چکیده
Machine Learning (ML) algorithms, especially deep neural networks (DNN), have proven themselves to be extremely useful tools for data analysis, and are increasingly being deployed in systems operating on sensitive data, such as recommendation systems, banking fraud detection, healthcare systems. This underscores the need privacy-preserving ML (PPML) has inspired a line of research into how can constructed efficiently. However, most prior works PPML achieve efficiency by requiring advanced algorithms simplified or substituted with approximated variants that “MPC-friendly” before multi-party computation (MPC) techniques applied obtain A drawback this approach is it requires careful fine-tuning combined MPC might lead less efficient inferior quality (such lower prediction accuracy). an issue secure training DNNs particular, involves several arithmetic thought “MPCunfriendly”, namely, integer division, exponentiation, inversion, square root extraction. In work, we take structurally different propose framework allows evaluation full-fledged state-of-the-art via computation. Specifically, protocols above seemingly MPC-unfriendly computations (but which essential DNN). Our three-party honest-majority setting, both passively actively abort variants. notable feature our they simultaneously provide high accuracy efficiency. enables us efficiently securely compute modern Adam (Adaptive moment estimation) softmax function “as is”, without resorting approximations. As result, DNN outperforms threeparty systems; full up 6.7 times faster than just online phase FALCON (Wagh et al. at PETS’21) 4.2 Dalskov (USENIX’21) standard benchmark network DNNs. The potential advantage even greater when considering more complex realistic networks. To demonstrate this, perform measurements real-world DNNs, AlexNet VGG16, large containing millions parameters. performance these factor 26 ∼ 33 48 51 VGG16 60% 70%, respectively, compared FALCON. Even CRYPTGPU (Tan IEEE S&P’21), optimized runs powerful GPUs, achieves 2.1 4.1 performance,
منابع مشابه
Adaptive dropout for training deep neural networks
Recently, it was shown that deep neural networks can perform very well if the activities of hidden units are regularized during learning, e.g, by randomly dropping out 50% of their activities. We describe a method called ‘standout’ in which a binary belief network is overlaid on a neural network and is used to regularize of its hidden units by selectively setting activities to zero. This ‘adapt...
متن کاملEmbedding-Based Speaker Adaptive Training of Deep Neural Networks
An embedding-based speaker adaptive training (SAT) approach is proposed and investigated in this paper for deep neural network acoustic modeling. In this approach, speaker embedding vectors, which are a constant given a particular speaker, are mapped through a control network to layer-dependent elementwise affine transformations to canonicalize the internal feature representations at the output...
متن کاملAdaptive Normalized Risk-Averting Training for Deep Neural Networks
This paper proposes a set of new error criteria and learning approaches, Adaptive Normalized Risk-Averting Training (ANRAT), to attack the non-convex optimization problem in training deep neural networks (DNNs). Theoretically, we demonstrate its effectiveness on global and local convexity lower-bounded by the standard Lp-norm error. By analyzing the gradient on the convexity index λ, we explain...
متن کاملAdaBatch: Adaptive Batch Sizes for Training Deep Neural Networks
Training deep neural networks with Stochastic Gradient Descent, or its variants, requires careful choice of both learning rate and batch size. While smaller batch sizes generally converge in fewer training epochs, larger batch sizes offer more parallelism and hence better computational efficiency. We have developed a new training approach that, rather than statically choosing a single batch siz...
متن کاملFast and Accurate Inference with Adaptive Ensemble Prediction in Image Classification with Deep Neural Networks
Ensembling multiple predictions is a widely used technique to improve the accuracy of various machine learning tasks. In image classification tasks, for example, averaging the predictions for multiple patches extracted from the input image significantly improves accuracy. Using multiple networks trained independently to make predictions improves accuracy further. One obvious drawback of the ens...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Proceedings on Privacy Enhancing Technologies
سال: 2022
ISSN: ['2299-0984']
DOI: https://doi.org/10.56553/popets-2022-0131